Picture for Zechun Liu

Zechun Liu

R-Sparse: Rank-Aware Activation Sparsity for Efficient LLM Inference

Add code
Apr 28, 2025
Viaarxiv icon

Param$Δ$ for Direct Weight Mixing: Post-Train Large Language Model at Zero Cost

Add code
Apr 23, 2025
Viaarxiv icon

PARQ: Piecewise-Affine Regularized Quantization

Add code
Mar 19, 2025
Viaarxiv icon

ParetoQ: Scaling Laws in Extremely Low-bit LLM Quantization

Add code
Feb 04, 2025
Viaarxiv icon

Efficient Track Anything

Add code
Nov 28, 2024
Viaarxiv icon

Llama Guard 3-1B-INT4: Compact and Efficient Safeguard for Human-AI Conversations

Add code
Nov 18, 2024
Viaarxiv icon

LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding

Add code
Oct 22, 2024
Figure 1 for LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding
Figure 2 for LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding
Figure 3 for LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding
Figure 4 for LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding
Viaarxiv icon

Agent-as-a-Judge: Evaluate Agents with Agents

Add code
Oct 14, 2024
Figure 1 for Agent-as-a-Judge: Evaluate Agents with Agents
Figure 2 for Agent-as-a-Judge: Evaluate Agents with Agents
Figure 3 for Agent-as-a-Judge: Evaluate Agents with Agents
Figure 4 for Agent-as-a-Judge: Evaluate Agents with Agents
Viaarxiv icon

Scaling Parameter-Constrained Language Models with Quality Data

Add code
Oct 04, 2024
Figure 1 for Scaling Parameter-Constrained Language Models with Quality Data
Figure 2 for Scaling Parameter-Constrained Language Models with Quality Data
Figure 3 for Scaling Parameter-Constrained Language Models with Quality Data
Figure 4 for Scaling Parameter-Constrained Language Models with Quality Data
Viaarxiv icon

RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization

Add code
Jul 10, 2024
Figure 1 for RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization
Figure 2 for RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization
Figure 3 for RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization
Figure 4 for RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization
Viaarxiv icon